multi-label classifier
- North America > United States > Illinois > Champaign County > Urbana (0.04)
- Asia > Singapore (0.04)
- Information Technology > Security & Privacy (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.69)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Optimization (0.47)
- Europe > Poland > Greater Poland Province > Poznań (0.04)
- Oceania > Australia > Victoria > Melbourne (0.04)
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- (2 more...)
MultiGuard: Provably Robust Multi-label Classification against Adversarial Examples
Multi-label classification, which predicts a set of labels for an input, has many applications. However, multiple recent studies showed that multi-label classification is vulnerable to adversarial examples. In particular, an attacker can manipulate the labels predicted by a multi-label classifier for an input via adding carefully crafted, human-imperceptible perturbation to it. Existing provable defenses for multi-class classification achieve sub-optimal provable robustness guarantees when generalized to multi-label classification. In this work, we propose MultiGuard, the first provably robust defense against adversarial examples to multi-label classification. Our MultiGuard leverages randomized smoothing, which is the state-of-the-art technique to build provably robust classifiers.
- Europe > Poland > Greater Poland Province > Poznań (0.04)
- Oceania > Australia > Victoria > Melbourne (0.04)
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- (2 more...)
Least-Ambiguous Multi-Label Classifier
Hagos, Misgina Tsighe, Lundström, Claes
Abstract--Multi-label learning often requires identifying all relevant labels for training instances, but collecting full label annotations is costly and labor-intensive. In many datasets, only a single positive label is annotated per training instance, despite the presence of multiple relevant labels. This setting, known as single-positive multi-label learning (SPMLL), presents a significant challenge due to its extreme form of partial supervision. We propose a model-agnostic approach to SPMLL that draws on conformal prediction to produce calibrated set-valued outputs, enabling reliable multi-label predictions at test time. We evaluate our approach on 12 benchmark datasets, demonstrating consistent improvements over existing baselines and practical applicability.
- North America > United States > Illinois > Champaign County > Urbana (0.04)
- Asia > Singapore (0.04)
PatchDEMUX: A Certifiably Robust Framework for Multi-label Classifiers Against Adversarial Patches
Jacob, Dennis, Xiang, Chong, Mittal, Prateek
Deep learning techniques have enabled vast improvements in computer vision technologies. Nevertheless, these models are vulnerable to adversarial patch attacks which catastrophically impair performance. The physically realizable nature of these attacks calls for certifiable defenses, which feature provable guarantees on robustness. While certifiable defenses have been successfully applied to single-label classification, limited work has been done for multi-label classification. In this work, we present PatchDEMUX, a certifiably robust framework for multi-label classifiers against adversarial patches. Our approach is a generalizable method which can extend any existing certifiable defense for single-label classification; this is done by considering the multi-label classification task as a series of isolated binary classification problems to provably guarantee robustness. Furthermore, in the scenario where an attacker is limited to a single patch we propose an additional certification procedure that can provide tighter robustness bounds. Using the current state-of-the-art (SOTA) single-label certifiable defense PatchCleanser as a backbone, we find that PatchDEMUX can achieve non-trivial robustness on the MS-COCO and PASCAL VOC datasets while maintaining high clean performance
- North America > United States > California > Alameda County > Berkeley (0.40)
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- North America > United States > New Jersey > Mercer County > Princeton (0.04)
- North America > United States > Louisiana > Orleans Parish > New Orleans (0.04)
- Information Technology > Security & Privacy (0.68)
- Education > Educational Setting > Higher Education (0.40)
MultiGuard: Provably Robust Multi-label Classification against Adversarial Examples
Multi-label classification, which predicts a set of labels for an input, has many applications. However, multiple recent studies showed that multi-label classification is vulnerable to adversarial examples. In particular, an attacker can manipulate the labels predicted by a multi-label classifier for an input via adding carefully crafted, human-imperceptible perturbation to it. Existing provable defenses for multi-class classification achieve sub-optimal provable robustness guarantees when generalized to multi-label classification. In this work, we propose MultiGuard, the first provably robust defense against adversarial examples to multi-label classification. Our MultiGuard leverages randomized smoothing, which is the state-of-the-art technique to build provably robust classifiers.
AEMLO: AutoEncoder-Guided Multi-Label Oversampling
Zhou, Ao, Liu, Bin, Wang, Jin, Sun, Kaiwei, Liu, Kelin
Oversampling is one of the most popular approaches, as it augments instances associated with less frequent labels to balance the class distribution. Existing oversampling methods generate feature vectors of synthetic samples through replication or linear interpolation and assign labels through neighborhood information. Linear interpolation typically generates new samples between existing data points, which may result in insufficient diversity of synthesized samples and further lead to the overfitting issue. Deep learning-based methods, such as AutoEncoders, have been proposed to generate more diverse and complex synthetic samples, achieving excellent performance on imbalanced binary or multi-class datasets. In this study, we introduce AEMLO, an AutoEncoder-guided Oversampling technique specifically designed for tackling imbalanced multi-label data. AEMLO is built upon two fundamental components. The first is an encoder-decoder architecture that enables the model to encode input data into a low-dimensional feature space, learn its latent representations, and then reconstruct it back to its original dimension, thus applying to the generation of new data. The second is an objective function tailored to optimize the sampling task for multi-label scenarios. We show that AEMLO outperforms the existing state-of-the-art methods with extensive empirical studies.
- Asia > China > Chongqing Province > Chongqing (0.04)
- Europe > Slovenia > Upper Carniola > Municipality of Bled > Bled (0.04)
- Europe > Poland > Masovia Province > Warsaw (0.04)
- Research Report > Promising Solution (0.34)
- Research Report > New Finding (0.34)
DALLMi: Domain Adaption for LLM-based Multi-label Classifier
Beţianu, Miruna, Mălan, Abele, Aldinucci, Marco, Birke, Robert, Chen, Lydia
Large language models (LLMs) increasingly serve as the backbone for classifying text associated with distinct domains and simultaneously several labels (classes). When encountering domain shifts, e.g., classifier of movie reviews from IMDb to Rotten Tomatoes, adapting such an LLM-based multi-label classifier is challenging due to incomplete label sets at the target domain and daunting training overhead. The existing domain adaptation methods address either image multi-label classifiers or text binary classifiers. In this paper, we design DALLMi, Domain Adaptation Large Language Model interpolator, a first-of-its-kind semi-supervised domain adaptation method for text data models based on LLMs, specifically BERT. The core of DALLMi is the novel variation loss and MixUp regularization, which jointly leverage the limited positively labeled and large quantity of unlabeled text and, importantly, their interpolation from the BERT word embeddings. DALLMi also introduces a label-balanced sampling strategy to overcome the imbalance between labeled and unlabeled data. We evaluate DALLMi against the partial-supervised and unsupervised approach on three datasets under different scenarios of label availability for the target domain. Our results show that DALLMi achieves higher mAP than unsupervised and partially-supervised approaches by 19.9% and 52.2%, respectively.
- Europe > Italy > Piedmont > Turin Province > Turin (0.14)
- Europe > Switzerland > Neuchâtel > Neuchâtel (0.04)
- Europe > Netherlands > South Holland > Delft (0.04)
- (2 more...)
- Media > Film (0.67)
- Leisure & Entertainment (0.49)